Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher.
Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?
Some links on this page may take you to non-federal websites. Their policies may differ from this site.
-
Background Pituitary adenylate cyclase-activating polypeptide (PACAP) is a multipotent neuropeptide widely distributed in the trigeminovascular system (TVS) and higher brain regions. At present, the underlying mechanism of PACAP/PACAP type1 (PAC1) receptor in migraine generation remains unclear. Methods The rat model of chronic migraine (CM) was established by repeated intraperitoneal injection of nitroglycerin (NTG). Von Frey filaments and hot plate tests were used to measure the mechanical and thermal thresholds. The expression levels of c-Fos, calcitonin gene-related peptide (CGRP), PACAP, PAC1, protein kinase A (PKA) and phosphorylated extracellular signal-regulated kinase (ERK) were assessed by western blotting or immunofluorescence staining. The internalization of PAC1 receptor was visualized by fluorescence microscope and laser scanning confocal microscope. Results The results showed that c-Fos and CGRP expression significantly increased after repeated administrations of NTG or PACAP. Pitstop2 notably improved hyperalgesia in CM rats, while PACAP6–38 offered no benefit. In addition, PACAP-induced PAC1 receptor internalization, PKA and ERK pathways activation were blocked by Pitstop2 instead of PACAP6–38. Conclusions Our results demonstrate that inhibition of PAC1 receptor internalization could effectively improve allodynia in CM rats by restraining ERK signaling pathway activation in a chronic migraine rat model. Modulation of receptor internalization may be a novel perspective to explore specific mechanisms of PACAP signaling activation in the trigeminal vascular system.more » « less
-
In many prediction problems, spurious correlations are induced by a changing relationship between the label and a nuisance variable that is also correlated with the covariates. For example, in classifying animals in natural images, the background, which is a nuisance, can predict the type of animal. This nuisance-label relationship does not always hold, and the performance of a model trained under one such relationship may be poor on data with a different nuisance-label relationship. To build predictive models that perform well regardless of the nuisance-label relationship, we develop Nuisance-Randomized Distillation (NURD). We introduce the nuisance-randomized distribution, a distribution where the nuisance and the label are independent. Under this distribution, we define the set of representations such that conditioning on any member, the nuisance and the label remain independent. We prove that the representations in this set always perform better than chance, while representations outside of this set may not. NURD finds a representation from this set that is most informative of the label under the nuisance-randomized distribution, and we prove that this representation achieves the highest performance regardless of the nuisance-label relationship. We evaluate NURD on several tasks including chest X-ray classification where, using non-lung patches as the nuisance, NURD produces models that predict pneumonia under strong spurious correlations.more » « less
-
Comparing the inferences of diverse candidate models is an essential part of model checking and escaping local optima. To enable efficient comparison, we introduce an amortized variational inference framework that can perform fast and reliable posterior estimation across models of the same architecture. Our Any Parameter Encoder (APE) extends the encoder neural network common in amortized inference to take both a data feature vector and a model parameter vector as input. APE thus reduces posterior inference across unseen data and models to a single forward pass. In experiments comparing candidate topic models for synthetic data and product reviews, our Any Parameter Encoder yields comparable posteriors to more expensive methods in far less time, especially when the encoder architecture is designed in model-aware fashion.more » « less
-
Comparing the inferences of diverse candidate models is an essential part of model checking and escaping local optima. To enable efficient comparison, we introduce an amortized variational inference framework that can perform fast and reliable posterior estimation across models of the same architecture. Our Any Parameter Encoder (APE) extends the encoder neural network common in amortized inference to take both a data feature vector and a model parameter vector as input. APE thus reduces posterior inference across unseen data and models to a single forward pass. In experiments comparing candidate topic models for synthetic data and product reviews, our Any Parameter Encoder yields comparable posteriors to more expensive methods in far less time, especially when the encoder architecture is designed in model-aware fashion.more » « less
An official website of the United States government

Full Text Available